Skip to content

Commit 5ec2829

Browse files
eu9eneCircleCI evaluation job
and
CircleCI evaluation job
authored
Add English to Dutch (#59)
* Add English to Dutch * Update evaluation results [skip ci] * Update model registry [skip ci] Co-authored-by: CircleCI evaluation job <ci-models-evaluation@firefox-translations>
1 parent 46f6780 commit 5ec2829

15 files changed

+57
-8
lines changed

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,6 +90,7 @@ Suffix of the model file in the registry:
9090
- Icelandic -> English
9191
- Norwegian Nynorsk -> English
9292
- Ukrainian <-> English
93+
- Dutch <- English
9394

9495
## Upcoming
95-
- Dutch <-> English
96+
- Dutch -> English
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
27.6
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
29.4
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
29.3
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
27.2
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
29.1
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
28.6

evaluation/dev/img/avg.png

345 Bytes
Loading

evaluation/dev/img/en-nl.png

20.2 KB
Loading

evaluation/dev/results.md

Lines changed: 15 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -57,11 +57,11 @@ Both absolute and relative differences in BLEU scores between Bergamot and other
5757

5858
## avg
5959

60-
| Translator/Dataset | en-uk | fa-en | ru-en | en-ru | uk-en | en-fa | is-en |
61-
| --- | --- | --- | --- | --- | --- | --- | --- |
62-
| bergamot | 28.00 | 28.70 | 33.37 | 30.47 | 35.65 | 17.30 | 23.50 |
63-
| google | 32.40 (+4.40, +15.71%) | 36.05 (+7.35, +25.61%) | 36.53 (+3.15, +9.45%) | 33.72 (+3.25, +10.67%) | 38.90 (+3.25, +9.12%) | 27.70 (+10.40, +60.12%) | 34.95 (+11.45, +48.72%) |
64-
| microsoft | 31.05 (+3.05, +10.89%) | 36.15 (+7.45, +25.96%) | 36.87 (+3.50, +10.49%) | 33.68 (+3.21, +10.53%) | 39.00 (+3.35, +9.40%) | 20.50 (+3.20, +18.50%) | 34.90 (+11.40, +48.51%) |
60+
| Translator/Dataset | en-uk | fa-en | ru-en | en-ru | uk-en | en-fa | en-nl | is-en |
61+
| --- | --- | --- | --- | --- | --- | --- | --- | --- |
62+
| bergamot | 28.00 | 28.70 | 33.37 | 30.47 | 35.65 | 17.30 | 27.40 | 23.50 |
63+
| google | 32.40 (+4.40, +15.71%) | 36.05 (+7.35, +25.61%) | 36.53 (+3.15, +9.45%) | 33.72 (+3.25, +10.67%) | 38.90 (+3.25, +9.12%) | 27.70 (+10.40, +60.12%) | 29.25 (+1.85, +6.75%) | 34.95 (+11.45, +48.72%) |
64+
| microsoft | 31.05 (+3.05, +10.89%) | 36.15 (+7.45, +25.96%) | 36.87 (+3.50, +10.49%) | 33.68 (+3.21, +10.53%) | 39.00 (+3.35, +9.40%) | 20.50 (+3.20, +18.50%) | 28.95 (+1.55, +5.66%) | 34.90 (+11.40, +48.51%) |
6565

6666
![Results](img/avg.png)
6767

@@ -125,6 +125,16 @@ Both absolute and relative differences in BLEU scores between Bergamot and other
125125

126126
![Results](img/en-fa.png)
127127

128+
## en-nl
129+
130+
| Translator/Dataset | flores-dev | flores-test |
131+
| --- | --- | --- |
132+
| bergamot | 27.60 | 27.20 |
133+
| google | 29.40 (+1.80, +6.52%) | 29.10 (+1.90, +6.99%) |
134+
| microsoft | 29.30 (+1.70, +6.16%) | 28.60 (+1.40, +5.15%) |
135+
136+
![Results](img/en-nl.png)
137+
128138
## is-en
129139

130140
| Translator/Dataset | flores-dev | flores-test |

evaluation/prod/results.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ BLEU Score | Interpretation
2424
Source: https://cloud.google.com/translate/automl/docs/evaluate#bleu
2525

2626

27-
BLEU is the most popular becnhmark in academia, so using BLEU allows us also to compare with reserach papers results and competitions (see [Conference on Machine Translation (WMT)](http://statmt.org/wmt21/)).
27+
BLEU is the most popular becnhmark in academia, so using BLEU allows us also to compare with reserach papers results and competitions (see [Conference on Machine Translation Conference (WMT)](http://statmt.org/wmt21/)).
2828

2929
Read [this article](https://www.rws.com/blog/understanding-mt-quality-bleu-scores/) to better understand what BLEU is and why it is not perfect.
3030

@@ -253,4 +253,4 @@ Both absolute and relative differences in BLEU scores between Bergamot and other
253253
| google | 24.70 (+0.40, +1.65%) | 38.60 (-1.40, -3.50%) | 24.10 (+0.70, +2.99%) | 33.70 (+0.60, +1.81%) | 28.80 (+0.60, +2.13%) | 28.90 (+2.20, +8.24%) | 23.70 (+0.10, +0.42%) | 26.50 (-0.30, -1.12%) | 43.50 (-1.00, -2.25%) | 30.90 (+1.10, +3.69%) | 36.50 (+0.80, +2.24%) | 42.30 (+3.50, +9.02%) | 47.80 (+0.10, +0.21%) | 31.50 (-0.50, -1.56%) | 23.60 (+0.60, +2.61%) | 43.70 (+4.90, +12.63%) |
254254
| microsoft | 25.30 (+1.00, +4.12%) | 40.50 (+0.50, +1.25%) | 23.70 (+0.30, +1.28%) | 34.30 (+1.20, +3.63%) | 28.80 (+0.60, +2.13%) | 28.20 (+1.50, +5.62%) | 24.00 (+0.40, +1.69%) | 27.20 (+0.40, +1.49%) | 43.80 (-0.70, -1.57%) | 32.20 (+2.40, +8.05%) | 36.10 (+0.40, +1.12%) | 42.90 (+4.10, +10.57%) | 48.70 (+1.00, +2.10%) | 33.10 (+1.10, +3.44%) | 23.90 (+0.90, +3.91%) | 44.00 (+5.20, +13.40%) |
255255

256-
![Results](img/en-de.png)
256+
![Results](img/en-de.png)
Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
version https://git-lfs.github.com/spec/v1
2+
oid sha256:f6798a31ddc076cf66909920297916c674d7b1e2866e3cbc79066b94d687f54f
3+
size 2454349
Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
version https://git-lfs.github.com/spec/v1
2+
oid sha256:59afabea0afef874c640964cfba0bfac3f1219894df01fa0603eb2acd81b4637
3+
size 13081379

models/dev/ennl/vocab.ennl.spm.gz

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
version https://git-lfs.github.com/spec/v1
2+
oid sha256:eef8a7d0a3275cce8f04496b7f7cb6686c52b6ded490edf5bc9a682b1d6e9a6d
3+
size 411799

registry.json

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -480,6 +480,29 @@
480480
"modelType": "dev"
481481
}
482482
},
483+
"ennl": {
484+
"model": {
485+
"name": "model.ennl.intgemm.alphas.bin",
486+
"size": 17140899,
487+
"estimatedCompressedSize": 13081379,
488+
"expectedSha256Hash": "906690a58a0d72aff28bd4b941cbd0984d1e0a62958c0b21aebae378a656d822",
489+
"modelType": "dev"
490+
},
491+
"lex": {
492+
"name": "lex.50.50.ennl.s2t.bin",
493+
"size": 4494892,
494+
"estimatedCompressedSize": 2454349,
495+
"expectedSha256Hash": "f780a6d74af4b141f551dcc0da56bab44a05a90ef53d63381269710f35eaa41b",
496+
"modelType": "dev"
497+
},
498+
"vocab": {
499+
"name": "vocab.ennl.spm",
500+
"size": 807541,
501+
"estimatedCompressedSize": 411799,
502+
"expectedSha256Hash": "43ba3922c3bba2b76ca2e2124837c96518b0e31300b7d6d5ccce55ee10d86393",
503+
"modelType": "dev"
504+
}
505+
},
483506
"enru": {
484507
"model": {
485508
"name": "model.enru.intgemm.alphas.bin",

0 commit comments

Comments
 (0)